Goto

Collaborating Authors

 neurips 2022


Generalized Laplacian Eigenmaps

Neural Information Processing Systems

Graph contrastive learning attracts/disperses node representations for similar/dissimilar node pairs under some notion of similarity. It may be combined with a low-dimensional embedding of nodes to preserve intrinsic and structural properties of a graph. COLES, a recent graph contrastive method combines traditional graph embedding and negative sampling into one framework. COLES in fact minimizes the trace difference between the within-class scatter matrix encapsulating the graph connectivity and the total scatter matrix encapsulating negative sampling. In this paper, we propose a more essential framework for graph embedding, called Generalized Laplacian EigeNmaps (GLEN), which learns a graph representation by maximizing the rank difference between the total scatter matrix and the within-class scatter matrix, resulting in the minimum class separation guarantee. However, the rank difference minimization is an NP-hard problem. Thus, we replace the trace difference that corresponds to the difference of nuclear norms by the difference of LogDet expressions, which we argue is a more accurate surrogate for the NP-hard rank difference than the trace difference. While enjoying a lesser computational cost, the difference of LogDet terms is lower-bounded by the Affine-invariant Riemannian metric (AIRM) and Jesen-Bregman the LogDet Divergence (JBLD), and upper-bounded by AIRM scaled by the factor of $\sqrt{m}$. We show that GLEN offers favourable accuracy/scalability compared to state-of-the-art baselines.


The NeurIPS 2022 Neural MMO Challenge: A Massively Multiagent Competition with Specialization and Trade

Liu, Enhong, Suarez, Joseph, You, Chenhui, Wu, Bo, Chen, Bingcheng, Hu, Jun, Chen, Jiaxin, Zhu, Xiaolong, Zhu, Clare, Togelius, Julian, Mohanty, Sharada, Hong, Weijun, Du, Rui, Zhang, Yibing, Wang, Qinwen, Li, Xinhang, Yuan, Zheng, Li, Xiang, Huang, Yuejia, Zhang, Kun, Yang, Hanhui, Tang, Shiqi, Isola, Phillip

arXiv.org Artificial Intelligence

In this paper, we present the results of the NeurIPS-2022 Neural MMO Challenge, which attracted 500 participants and received over 1,600 submissions. Like the previous IJCAI-2022 Neural MMO Challenge, it involved agents from 16 populations surviving in procedurally generated worlds by collecting resources and defeating opponents. This year's competition runs on the latest v1.6 Neural MMO, which introduces new equipment, combat, trading, and a better scoring system. These elements combine to pose additional robustness and generalization challenges not present in previous competitions. This paper summarizes the design and results of the challenge, explores the potential of this environment as a benchmark for learning methods, and presents some practical reinforcement learning training approaches for complex tasks with sparse rewards. Additionally, we have open-sourced our baselines, including environment wrappers, benchmarks, and visualization tools for future research.


An ensemble of VisNet, Transformer-M, and pretraining models for molecular property prediction in OGB Large-Scale Challenge @ NeurIPS 2022

Wang, Yusong, Li, Shaoning, Wang, Zun, He, Xinheng, Shao, Bin, Liu, Tie-Yan, Wang, Tong

arXiv.org Artificial Intelligence

In the technical report, we provide our solution for OGB-LSC 2022 Graph Regression Task. The target of this task is to predict the quantum chemical property, HOMO-LUMO gap for a given molecule on PCQM4Mv2 dataset. In the competition, we designed two kinds of models: Transformer-M-ViSNet which is an geometry-enhanced graph neural network for fully connected molecular graphs and Pretrained-3D-ViSNet which is a pretrained ViSNet by distilling geomeotric information from optimized structures. With an ensemble of 22 models, ViSNet Team achieved the MAE of 0.0723 eV on the test-challenge set, dramatically reducing the error by 39.75% compared with the best method in the last year competition.


A Conceptual Model for End-to-End Causal Discovery in Knowledge Tracing

Kumar, Nischal Ashok, Feng, Wanyong, Lee, Jaewook, McNichols, Hunter, Ghosh, Aritra, Lan, Andrew

arXiv.org Artificial Intelligence

In this paper, we take a preliminary step towards solving the problem of causal discovery in knowledge tracing, i.e., finding the underlying causal relationship among different skills from real-world student response data. This problem is important since it can potentially help us understand the causal relationship between different skills without extensive A/B testing, which can potentially help educators to design better curricula according to skill prerequisite information. Specifically, we propose a conceptual solution, a novel causal gated recurrent unit (GRU) module in a modified deep knowledge tracing model, which uses i) a learnable permutation matrix for causal ordering among skills and ii) an optionally learnable lower-triangular matrix for causal structure among skills. We also detail how to learn the model parameters in an end-to-end, differentiable way. Our solution placed among the top entries in Task 3 of the NeurIPS 2022 Challenge on Causal Insights for Learning Paths in Education. We detail preliminary experiments as evaluated on the challenge's public leaderboard since the ground truth causal structure has not been publicly released, making detailed local evaluation impossible.


Traffic4cast at NeurIPS 2022 -- Predict Dynamics along Graph Edges from Sparse Node Data: Whole City Traffic and ETA from Stationary Vehicle Detectors

Neun, Moritz, Eichenberger, Christian, Martin, Henry, Spanring, Markus, Siripurapu, Rahul, Springer, Daniel, Deng, Leyan, Wu, Chenwang, Lian, Defu, Zhou, Min, Lumiste, Martin, Ilie, Andrei, Wu, Xinhua, Lyu, Cheng, Lu, Qing-Long, Mahajan, Vishal, Lu, Yichao, Li, Jiezhang, Li, Junjun, Gong, Yue-Jiao, Grötschla, Florian, Mathys, Joël, Wei, Ye, Haitao, He, Fang, Hui, Malm, Kevin, Tang, Fei, Kopp, Michael, Kreil, David, Hochreiter, Sepp

arXiv.org Artificial Intelligence

The global trends of urbanization and increased personal mobility force us to rethink the way we live and use urban space. The Traffic4cast competition series tackles this problem in a data-driven way, advancing the latest methods in machine learning for modeling complex spatial systems over time. In this edition, our dynamic road graph data combine information from road maps, $10^{12}$ probe data points, and stationary vehicle detectors in three cities over the span of two years. While stationary vehicle detectors are the most accurate way to capture traffic volume, they are only available in few locations. Traffic4cast 2022 explores models that have the ability to generalize loosely related temporal vertex data on just a few nodes to predict dynamic future traffic states on the edges of the entire road graph. In the core challenge, participants are invited to predict the likelihoods of three congestion classes derived from the speed levels in the GPS data for the entire road graph in three cities 15 min into the future. We only provide vehicle count data from spatially sparse stationary vehicle detectors in these three cities as model input for this task. The data are aggregated in 15 min time bins for one hour prior to the prediction time. For the extended challenge, participants are tasked to predict the average travel times on super-segments 15 min into the future - super-segments are longer sequences of road segments in the graph. The competition results provide an important advance in the prediction of complex city-wide traffic states just from publicly available sparse vehicle data and without the need for large amounts of real-time floating vehicle data.


Time-Myopic Go-Explore: Learning A State Representation for the Go-Explore Paradigm

Höftmann, Marc, Robine, Jan, Harmeling, Stefan

arXiv.org Artificial Intelligence

Very large state spaces with a sparse reward signal are difficult to explore. The lack of a sophisticated guidance results in a poor performance for numerous reinforcement learning algorithms. In these cases, the commonly used random exploration is often not helpful. The literature shows that this kind of environments require enormous efforts to systematically explore large chunks of the state space. Learned state representations can help here to improve the search by providing semantic context and build a structure on top of the raw observations. In this work we introduce a novel time-myopic state representation that clusters temporal close states together while providing a time prediction capability between them. By adapting this model to the Go-Explore paradigm (Ecoffet et al., 2021b), we demonstrate the first learned state representation that reliably estimates novelty instead of using the hand-crafted representation heuristic. Our method shows an improved solution for the detachment problem which still remains an issue at the Go-Explore Exploration Phase. We provide evidence that our proposed method covers the entire state space with respect to all possible time trajectories without causing disadvantageous conflict-overlaps in the cell archive. Analogous to native Go-Explore, our approach is evaluated on the hard exploration environments MontezumaRevenge, Gravitar and Frostbite (Atari) in order to validate its capabilities on difficult tasks. Our experiments show that time-myopic Go-Explore is an effective alternative for the domain-engineered heuristic while also being more general. The source code of the method is available on GitHub.


Emergent collective intelligence from massive-agent cooperation and competition

Chen, Hanmo, Tao, Stone, Chen, Jiaxin, Shen, Weihan, Li, Xihui, Yu, Chenghui, Cheng, Sikai, Zhu, Xiaolong, Li, Xiu

arXiv.org Artificial Intelligence

Inspired by organisms evolving through cooperation and competition between different populations on Earth, we study the emergence of artificial collective intelligence through massive-agent reinforcement learning. To this end, We propose a new massive-agent reinforcement learning environment, Lux, where dynamic and massive agents in two teams scramble for limited resources and fight off the darkness. In Lux, we build our agents through the standard reinforcement learning algorithm in curriculum learning phases and leverage centralized control via a pixel-to-pixel policy network. As agents co-evolve through self-play, we observe several stages of intelligence, from the acquisition of atomic skills to the development of group strategies. Since these learned group strategies arise from individual decisions without an explicit coordination mechanism, we claim that artificial collective intelligence emerges from massive-agent cooperation and competition. We further analyze the emergence of various learned strategies through metrics and ablation studies, aiming to provide insights for reinforcement learning implementations in massive-agent environments.


Affinity group round-up from NeurIPS 2022

AIHub

It was a busy month for affinity groups at NeurIPS, with workshops from Black in AI, Queer in AI, LatinX in AI, Indigenous in AI, Global South in AI, Women in ML, and North Africans in ML. These workshops give researchers the opportunity to share their work, find support and make connections, and raise awareness of issues affecting their communities. Here are some of our highlights from the workshops. David Adelani presented his work on transfer languages – taking a model in one language and applying it to other languages. Transferring from one to another language can be tricky, especially when they use different structures or scripts.



Nvidia showcases groundbreaking generative AI research at NeurIPS 2022

#artificialintelligence

Check out the on-demand sessions from the Low-Code/No-Code Summit to learn how to successfully innovate and achieve efficiency by upskilling and scaling citizen developers. The hardware giant continues to push the boundaries of technology in machine learning (ML), self-driving cars, robotics, graphics, simulation and more. The three categories of awards at NeurIPS 2022 were these: outstanding main track papers, outstanding datasets and benchmark track papers, and the test of time paper. Nvidia also presented a series of AI advancements it had worked on for the past year. It has released two papers, on providing unique lighting approaches and on 3D model creation, following up on its work in 3D and generative AI. "NeurIPS is a major conference in machine learning, and we see high value in participating in the show among other leaders in the field. We showcased 60 research projects at the conference and were proud to have two papers honored with NeurIPS 2022 Awards for their contributions to machine learning," Sanja Fidler, VP of AI research at Nvidia and a writer on both the 3D MoMa and GET3D papers, told VentureBeat.